Learning DNF by Approximating Inclusion-Exclusion Formulae

نویسندگان

  • Jun Tarui
  • Tatsuie Tsukiji
چکیده

Probably Approximately Correct learning algorithms generalize a small number of examples about an unknown concept into a function that can predict a future observation. More formally, let X and Y be the instance and outcome spaces, respectively. Then a PAC algorithm observes randomly drawn examples (x; f(x)) about an unknown concept f : X ! Y . These examples are independently and identically distributed random variables governed by an arbitrary and unknown distribution over X . With and only with these training examples, the algorithm aims to nd a hypothesis h : X ! Y that approximates the target concept f with respect to the same distribution. Hence it measures \goodness" of the hypothesis h by the probability acc(h) = Probx2Xfh(x) = f(x)g called the prediction accuracy. Valiant introduced the PAC model in a series of papers [12, 13], which is currently one of the most standard platforms for invention of polynomial-time learning algorithms. The PAC theory aims to learn as much general concept classes as possible, beginning from simple structures, e.g. depth-one or depth-two Boolean circuits. Valiant proved that Boolean conjunctions are polynomial-time learnable, and left the learning problem of the class DNF = fpolynomial-size Disjunctive Normal Form formulaeg for the future research. Here, as usual, a DNF formula is a disjunction of a amily of conjunctions of Boolean literals. These Boolean conjunctions are commonly called the terms of the DNF formula. The size of a DNF formula is the number of its (distinct) terms. Since then, a lot of literatures have proved learnability of subclasses of DNF by specifying either structural parameters of formulae or the distribution for the training examples ([1] provides a list of literatures).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Generalized Graph Colorability and Compressibility of Boolean Formulae

In this paper, we study the possibility of Occam’s razors for a widely studied class of Boolean Formulae : Disjunctive Normal Forms (DNF). An Occam’s razor is an algorithm which compresses the knowledge of observations (examples) in small formulae. We prove that approximating the minimally consistent DNF formula, and a generalization of graph colorability, is very hard. Our proof technique is s...

متن کامل

Learning DNF Expressions from Fourier Spectrum

Since its introduction by Valiant in 1984, PAC learning of DNF expressions remains one of the central problems in learning theory. We consider this problem in the setting where the underlying distribution is uniform, or more generally, a product distribution. Kalai, Samorodnitsky, and Teng (2009b) showed that in this setting a DNF expression can be efficiently approximated from its “heavy” low-...

متن کامل

On Learning Monotone DNF under Product Distributions

We show that the class of monotone 2O( √ log n)-term DNF formulae can be PAC learned in polynomial time under the uniform distribution from random examples only. This is an exponential improvement over the best previous polynomial-time algorithms in this model, which could learn monotone o(log n)-term DNF. We also show that various classes of small constant-depth circuits which compute monotone...

متن کامل

Quantum DNF Learnability Revisited

We describe a quantum PAC learning algorithm for DNF formulae under the uniform distribution with a query complexity of Õ(s/ǫ+ s/ǫ), where s is the size of DNF formula and ǫ is the PAC error accuracy. If s and 1/ǫ are comparable, this gives a modest improvement over a previously known classical query complexity of Õ(ns/ǫ). We also show a lower bound of Ω(s logn/n) on the query complexity of any...

متن کامل

ua nt - p h / 02 02 06 6 v 1 1 2 Fe b 20 02 Quantum DNF Learnability Revisited

We describe a quantum PAC learning algorithm for DNF formulae under the uniform distribution with a query complexity of Õ(s/ǫ+ s/ǫ), where s is the size of DNF formula and ǫ is the PAC error accuracy. If s and 1/ǫ are comparable, this gives a modest improvement over a previously known classical query complexity of Õ(ns/ǫ). We also show a lower bound of Ω(s logn/n) on the query complexity of any...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1999